Frequentist statistics are often the basic examples we learn in introductory statisitcs courses.
The philosophy of frequentism is effectively that there is some real truth out in the world about our data or process, and we trying to estimate the true parameters. We can increase our sample sizes, and run some analysis to try and determine what the truth is, and talk about thing like p-values and confidence limits to understand how far off reality we were.
Bayesians have a different intepretation of probability - we update our expectations based on the data, and try and form the most reasonable expectation.
A more obvious difference is the use of priors: In a bayesian framework we are interested in expectations: Expectations are modified by our prior beliefs.
Most Bayesians would argue that by default humans are Bayesian - no one would assume that the sun has gone nova based on the above evidence.
We incorporate evidence to change priors - in the case above, probably you would roll more dice, fund another particle detector, and call NASA before concluding that the sun is gone. In a less extreme case we probably hpld a weaker prior, and could be convinced by a few more rolls.
Let's take an example - We have a test for alcohol screening pilots on duty. We know that if a pilot has been drinking, 95% of the time, the test will be positive. Similarly, if the pilot has not been drinking, 99% of the time the test will be negative. We know that 0.5% of the pilot are drinking (this is our prior).
If we score a pilot as positive on the test, what is the chance that the pilot has been drinking?
In [2]:
### Show of hands for 95%, 50-95%, 10-50%, less than 10%
Bayes theorem, named after the Revered Thomas Bayes can be stated as below:
$$ P(A|B) = \frac{P(B|A)P(A)}{P(B)} $$In English, $P(A|B)$ means the probability of drinking (A), given a positive test (B).
$P(A|B)$ is the probability of a positive test, given a that the subject is drinking, $P(A)$ is the probability of drinking (prior).
Let's plug in our numbers:
In [3]:
puser = 0.005
pnonuser = 1 - puser
puserpos = 0.95
pnonuserpos = 1 - 0.99
prob = (puserpos * puser)/(puserpos*puser + pnonuserpos*pnonuser)
#prob
Unintuitive right?
So, we can see that with a strong prior, we cannot say much from a single test. Probably anyone who has taken an undergrad statistics course has seen a similar example - with the pilot changed to a criminal, or a medical test for a rare disease.
Then, Bayes is done for our class, and we move on. This is due to the fact that it is pretty difficult to do more without a lot of math.
How would we update our prior if we tested 100 people and saw a new ratio of drinkers? What if we retested the same pilot multiple times? These are solvable, but not easily without a lot more detail than most courses will want to give you.
Probabilistic programming allows us to get around a lot of the math.
Unlike the initial guess of what probabilistic programming does, it is not about writing code randomly - instead it means that we have probablitity distributions and model components in our programming language, that we can use to build models.
Let's put the above example into pymc3
to get a feel for the setup. It's a bit of overkill!
In [4]:
import pymc3 as pm
import numpy as np
###we use a context manager for the model:
with pm.Model() as drunk_pilots:
#prior
###a bernoulli is a single yes/no with a defined p
drinker = pm.Bernoulli('drinks', 0.005)
#testpos
#is a function of our drinker variable
test_pos = drinker * 0.95 + (1-drinker) * 0.01
##actualtest
##here we have the actual 'test'
##we can call this the likelihood function
##in this case we observed just 1 positive
tests = pm.Bernoulli('tests', test_pos, observed=1)
Our model is now defined!
Let's run it:
In [5]:
with drunk_pilots:
trace = pm.sample(draws=10000)
In [6]:
trace['drinks'].mean()
Out[6]:
In [7]:
prob
Out[7]:
Not a bad guess!
But, now imagine we want to update our model - we have been hired by an aircraft company to test 1000 of their pilots. They have a feeling that they might have a problem, so we want to be able to update our prior rate that was previously set at 1.
We can give our prior it's own prior! In this case, let's set it at a mean of 0.005, but with an sd of 0.05 - we allow our parameter to vary a little. I will choose a normal distribution for now - we could use any (or if we really wanted, use the conjugate prior).
We will also do some conversion of the model - a Bernoulli trial repeated N times is a binomial.
In [8]:
# we got 20 positives from 1000 pilots
data = np.append(np.ones(20),np.zeros(980))
N = len(data)
with pm.Model() as drunk_pilots:
#prior for p
drinksp = pm.Normal('drinksp',0.005, 0.1)
###a binomial is total number of drunks, with n positives
drinks = pm.Binomial('drinks', N, drinksp)
#testpos
#is a function of our drinks variable
test_pos = drinks/N * 0.95 + (1-drinks/N) * 0.01
##actualtest
##here we have the actual 'test'
##in this case we observed just 1 positive
tests = pm.Binomial('tests', N, test_pos, observed=data.sum())
In [10]:
with drunk_pilots:
trace = pm.sample(draws=10000)
pm.traceplot(trace)
Out[10]:
We can see we have substantially updated our prior - to
In [11]:
trace['drinksp'].mean()
Out[11]:
and have a total number of drinkers as
In [12]:
trace['drinks'].mean()
Out[12]:
We have just seen a small pymc3 model - how is it working under the hood?
PyMC3 was originally started as a project by Chris Fonnebeck, and was incorporated into numfocus last year (the organisation that umbrellas pyData, as well as numpy, pandas etc).
It is a package for Bayesian statistical modeling based on MCMC and variational inference.
Bayesian software has a relatively recent history: WinBugs, Stan, Edward and George as in the same area - in R Stan is the big competitor, and a few people might have memories of typing their own models into WinBugs.
Under the hood PyMC3 has some super cool stuff going on:
As a high level 'gist': when we make a model, we need to calculate the likelihood functions to sample. We then choose some random samples of our parameters, compare to the likelihoods, and add/remove them from the chain. We store the values in our chain, and call them our samples. We are using some smart sampling - we use NUTs, or Metropolis Hastings. This is a bit beyond this talk!
PyMC3 uses theano to calculate the computational graph of the model, and calculate the gradients - this means that our model is jit compiled into C, and is super quick. Unfortunately theano has reached end of life, so work is currently underway to port the backend to TensorFlow. By default, we have parallel processing - each core carries out one sample chain. We can use GPU if we want to set up theano.
In [13]:
pm.traceplot(trace)
Out[13]:
For the purposes of most people, we can decide that as long as we know the way to build a model, and can define the parameters and distributions correctly, the exact math doesn't work. Eric Ma has called the 'sample' function the 'Inference Button™', and this is how you can use it.
Now we have the basics, and an intuition for the workings, we can run more complicated models. A common method is to determine the difference in user response to websites: we know that we have x sales on site A, and y on site B, but how can we compare them? NB, this example is modified from the excellent text, 'Probabilistic-Programming-and-Bayesian-Methods-for-Hackers', by Cameron Davidson-Pilon.
Let's say we have set up an A/B test on our website:
We know our current website gets 30000 hits per day, and that it seems to work. We will pre-test the website with 5000 of those visitors sent to version 'B', leaving 25000 on 'A'.
We then want to compare if there is any difference in the sites:
In [15]:
from scipy import stats
np.random.seed(1234)
na = 25000
nb = 5000
p_a_buy = 0.04
p_b_buy = 0.05
data_a = stats.bernoulli.rvs(size = na, p = p_a_buy)
data_b = stats.bernoulli.rvs(size = nb, p = p_b_buy)
print(data_a.sum(), data_b.sum())
print(data_a.mean(), data_b.mean())
In frequentist stats, probably we would carry out a Fishers Exact or Chi Squared test - we create a contingency table of the data:
A | B | |
---|---|---|
bought | 1036 | 245 |
not bought | 23964 | 4755 |
And carry out the test:
In [16]:
oddsratio, pvalue = stats.fisher_exact([[1036, 245], [23964, 4755]])
oddsratio, pvalue
Out[16]:
So, we can see it is significant, but we don't really have a good idea of exactly what the effect is. What is the range of sales we can expect? How did it change?
Let's try it in pymc3:
In [17]:
with pm.Model() as abmodel:
###let's use uniform priors
###no idea how sales looks!
pa = pm.Uniform('pa', 0,1)
pb = pm.Uniform('pb', 0,1)
###we can create calculated values
###if they depend on others, and are not random
###use the deterministic function
diff = pm.Deterministic('diff', pa - pb)
###we have two obs, a and b
obsa = pm.Bernoulli("obsa", pa, observed=data_a)
obsb = pm.Bernoulli("obsb", pb, observed=data_b)
trace = pm.sample(20000)
burned_trace=trace[1000:]
In [18]:
pm.traceplot(burned_trace);
In [16]:
pm.posteriorplot.plot_posterior(burned_trace);
We get a distribution of the diffs. This is really nice - if we want to explore our expected sales, or if our redesign costs a lot more to run, we can know how much it might be worth.
We can also sample the posterior distribution of any of our observed samples - this is useful for diagnostics, but can also be used to simulate more data, given the calculated model:
In [17]:
ppc = pm.sample_ppc(trace, samples=500, model=abmodel, size=100)
In [18]:
print(np.mean(ppc['obsa']-ppc['obsb']))
We can continue on and make our models arbitraily complicated.
For this, let's model my commute (with a bit of exaggeration). Again, this is a modified example from Colin Carroll.
I have several ways of getting to work - and it will depend on various factors which I take, but they will add up to give an overall distribution of my commute time.
I leave my condo, and walk to the closest bixi stand. If there is a bike, I will bike to work, drop off the bixi, and I'm done. If there is no available bixi, I have the choice of taking a streetcar, or walking. Let's put this together into a model:
In [19]:
from theano import tensor as tt
with pm.Model() as commute_time:
#time to get to bixi
walk_to_bixi = pm.Normal('walk_to_bixi', 5, 1)
#chance of bixi being available
bixi_av = pm.Bernoulli('bixi_av', 0.2)
#Decision to take streetcar
take_streetcar = pm.Bernoulli('take_streetcar', 0.5)
#Streetcar time
streetcar_time = pm.Normal('streetcar_time', 15, 5)
#Walk time
walk_time = pm.Normal('walk_time', 30, 5)
#bixi time
bixi_time = pm.Normal('bixi_time', 7, 2)
###put it all together
total_time = pm.Deterministic('total_time',
walk_to_bixi +\
#if else!
tt.switch(bixi_av, bixi_time,\
tt.switch(take_streetcar,\
streetcar_time, walk_time)))
trace = pm.sample(size = 1000)
In [20]:
pm.traceplot(trace);
With a model like this, we can now seriously slice and dice the outcomes. What if I know that there is no bixi available?
In [21]:
from matplotlib import pyplot as plt
%matplotlib inline
plt.hist(trace['total_time'][trace['bixi_av'] == 0]);
How long will I commute if I took the streetcar?
In [22]:
plt.hist(trace['total_time'][trace['streetcar_time'] != 0]);
In frequentist terms, a linear regression can be seen as
$$ Y = X\beta + \epsilon$$Where Y is our dependent variable, X is our indepedent var, $\beta$ is a coefficient, and $\epsilon$ is some error term.
In Bayesian terms we can reformulate it as follows:
$$ Y \sim \mathcal{N}(X \beta, \sigma^2)$$To say that Y is a sample from a normal distribution with mean of $X\beta$ and sd of $\sigma^2$
This has some immediate cool effects:
We can use priors - if we think our sd is small, we can put a prior on it, rather than relying on the spread of our data.
We can create an entire distribution of $\beta$ based on our data, not a single point estimate.
Let's take a look:
In [23]:
from matplotlib import pyplot as plt
%matplotlib inline
n = 200
#we normally talk about y = ax + b
b = 5
a = 3
x = np.linspace(0, 1, n)
y = b + a * x + np.random.normal(scale=1, size=n)
plt.scatter(x,y);
#real line
plt.scatter(x, b + a * x);
Let's try some modeling:
In [24]:
with pm.Model() as lm:
# Define priors
sd = pm.HalfCauchy('sd', beta=10)
b = pm.Normal('b', 0, sd=20)
a = pm.Normal('a', 0, sd=20)
# Define likelihood
likelihood = pm.Normal('y', mu=b + a * x,
sd=sd, observed=y)
trace = pm.sample(1000)
In [25]:
pm.traceplot(trace);
In [26]:
trace['b'].mean(), trace['a'].mean(), trace['sd'].mean()
Out[26]:
In [27]:
import scipy
scipy.stats.linregress(x,y)
Out[27]:
Not bad! We can think a bit more about how we can change our priors, and create more complex models though....
There are a million other useful things we can do in PyMC3, not limited to the examples here:
The pymc docs are useful, albeit a little technical.
The authours maintain a binder of useful notebooks with interactive notebooks and examples.
Pycon Talks by Chris Fonnebeck and Eric Ma.
Bayesian Methods for Hackers is a great introduction to Bayes and PyMC3